Function level parallelism driven by data dependencies
نویسندگان
چکیده
منابع مشابه
Enhancing Parallelism by Removing Cyclic Data Dependencies
The parallel execution of loop iterations often is inhibited by recurrence relations on scalar variables. Examples are the use of induction variables and recursive functions. Due to the cyclic dependence between the iterations, these loops have to be executed sequentially. A method is presented to convert a family of coupled linear recurrence relations into explicit functions of a loop index. W...
متن کاملConstrained Data-Driven Parallelism
In data-driven parallelism, changes to data spawn new tasks, which may change more data, spawning yet more tasks. Computation propagates until no further changes occur. Benefits include increasing opportunities for finegrained parallelism, avoiding redundant work, and supporting incremental computations on large data sets. Nonetheless, data-driven parallelism can be problematic. For example, co...
متن کاملDetection of Function- level Parallelism
While the chipmultiprocessor (CMP) has quickly become the predominant processor architecture, its continuing success largely depends on the parallelizability of complex programs. We present a framework that is able to extract coarse-grain function-level parallelism that can exploit the parallel resources of the CMP. The framework uses a profile-driven control and data dependence analysis betwee...
متن کاملA Programming Model for Massive Data Parallelism with Data Dependencies
Accelerating processors can often be more cost and energy effective for a wide range of data-parallel computing problems than general-purpose processors. For graphics processor units (GPUs), this is particularly the case when program development is aided by environments, such as NVIDIA’s Compute Unified Device Architecture (CUDA), which dramatically reduces the gap between domainspecific archit...
متن کاملLimits of Data-Level Parallelism
A new breed of processors like the Cell Broadband Engine, the Imagine stream processor and the various GPU processors emphasize data-level parallelism (DLP) and threadlevel parallelism (TLP) as opposed to traditional instructionlevel parallelism (ILP). This allows them to achieve order-ofmagnitude improvements over conventional superscalar processors for many workloads. However, it is unclear a...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: ACM SIGARCH Computer Architecture News
سال: 2007
ISSN: 0163-5964
DOI: 10.1145/1241601.1241612